We address the challenge of speed and generalization in sketch-based 3D pose estimation by employing a learn-from-synthesis strategy. Through training on our established synthetic sketch-pose dataset, we present Sketch2PoseNet—a feed-forward network tailored to sketches—that efficiently and accurately predicts generalized 3D human poses across various sketch styles.

SIGGRAPH Asia 2025    Project Page    Code

We explore learning a native generative model for 360° full head from limited 3D head data. Three key problems are studied: 1) utilizing various representations for 360°-renderable head generation; 2) disentangling face appearance, shape, and motion for editable and motion-driven 3D head models; 3) enhancing model generalization for downstream tasks.
arXiv 2024

We present a novel approach for synthesizing 3D talking heads with controllable emotion, enhancing lip synchronization and rendering quality. To address multi-view consistency and emotional expressiveness issues, we propose a ‘Speech-to-Geometry-to-Appearance’ mapping framework trained on the EmoTalk3D dataset, enabling controllable emotion, wide-range view rendering, and fine facial details.
ECCV 2024    Project Page

We propose STAG4D, a novel framework for high-quality 4D generation, integrating pre-trained diffusion models with dynamic 3D Gaussian splatting. Our method outperforms prior 4D generation works in rendering quality, spatial-temporal consistency, and generation robustness, setting a new state-of-the-art for 4D generation from diverse inputs, including text, image, and video.
ECCV 2024    Project Page    Code

Author's picture

Hao Zhu

NJU-3DV Lab, Nanjing University
E-mail: zh@nju.edu.cn


Assistant Professor, PhD Advisor


Nanjing, China